2024-03-12 06:48:50
SCORE: Self-supervised Correspondence Fine-tuning for Improved Content Representations
Amit Meghanani, Thomas Hain
https://arxiv.org/abs/2403.06260 https:/…
SCORE: Self-supervised Correspondence Fine-tuning for Improved Content Representations
Amit Meghanani, Thomas Hain
https://arxiv.org/abs/2403.06260 https:/…
Structure-aware Fine-tuning for Code Pre-trained Models
Jiayi Wu, Renyu Zhu, Nuo Chen, Qiushi Sun, Xiang Li, Ming Gao
https://arxiv.org/abs/2404.07471 http…
Fine-tuning of diffusion models via stochastic control: entropy regularization and beyond
Wenpin Tang
https://arxiv.org/abs/2403.06279 https://
This https://arxiv.org/abs/2312.14378 has been replaced.
initial toot: https://mastoxiv.page/@arXiv_csLG_…
William Whitworth, who wrote revealing profiles in The New Yorker before working as the EIC at The Atlantic from 1980 to 1999, died on March 8 at age 87 (Sam Roberts/New York Times)
https://www.nytimes.com/2024/03/09/…
Delving into Parameter-Efficient Fine-Tuning in Code Change Learning: An Empirical Study
Shuo Liu, Jacky Keung, Zhen Yang, Fang Liu, Qilin Zhou, Yihan Liao
https://arxiv.org/abs/2402.06247
Adding NVMe SSDs to Enable and Accelerate 100B Model Fine-tuning on a Single GPU
Changyue Liao, Mo Sun, Zihan Yang, Kaiqi Chen, Binhang Yuan, Fei Wu, Zeke Wang
https://arxiv.org/abs/2403.06504
This https://arxiv.org/abs/2401.05126 has been replaced.
initial toot: https://mastoxiv.page/@arXiv_csCV_…
Improving Low-Resource Knowledge Tracing Tasks by Supervised Pre-training and Importance Mechanism Fine-tuning
Hengyuan Zhang, Zitao Liu, Shuyan Huang, Chenming Shang, Bojun Zhan, Yong Jiang
https://arxiv.org/abs/2403.06725
Fine-Tuning Surrogate Gradient Learning for Optimal Hardware Performance in Spiking Neural Networks
Ilkin Aliyev, Tosiron Adegbija
https://arxiv.org/abs/2402.06211
#ExplainVintageTechnology
#HashTagGames
It's a TV... but with no colour, the screen was curved, and you had to fiddle endlessly with the rabbit's ears until you finally got a good picture. The you had to twist a dial called "fine tuning" to try and get…
IndicLLMSuite: A Blueprint for Creating Pre-training and Fine-Tuning Datasets for Indian Languages
Mohammed Safi Ur Rahman Khan, Priyam Mehta, Ananth Sankar, Umashankar Kumaravelan, Sumanth Doddapaneni, Suriyaprasaad G, Varun Balan G, Sparsh Jain, Anoop Kunchukuttan, Pratyush Kumar, Raj Dabre, Mitesh M. Khapra
https://arxiv.org/…
Automated Federated Pipeline for Parameter-Efficient Fine-Tuning of Large Language Models
Zihan Fang, Zheng Lin, Zhe Chen, Xianhao Chen, Yue Gao, Yuguang Fang
https://arxiv.org/abs/2404.06448
Increased LLM Vulnerabilities from Fine-tuning and Quantization
Divyanshu Kumar, Anurakt Kumar, Sahil Agarwal, Prashanth Harshangi
https://arxiv.org/abs/2404.04392
Truthful Aggregation of LLMs with an Application to Online Advertising
Ermis Soumalias, Michael J. Curry, Sven Seuken
https://arxiv.org/abs/2405.05905 http…
Randomness Is All You Need: Semantic Traversal of Problem-Solution Spaces with Large Language Models
Thomas Sandholm, Sayandev Mukherjee, Bernardo A. Huberman
https://arxiv.org/abs/2402.06053
This https://arxiv.org/abs/2312.15698 has been replaced.
link: https://scholar.google.com/scholar?q=a
This https://arxiv.org/abs/2309.12307 has been replaced.
initial toot: https://mastoxiv.page/@arXiv_csCL_…
Si Superstrate Lenses on Patch-Antenna-Coupled TeraFETs: NEP Optimization and Frequency Fine-Tuning
Anastasiya Krysl, Dmytro B. But, K\k{e}stutis Ikamas, Jakob Holstein, Anna Shevchik-Shekera, Hartmut G. Roskos, Alvydas Lisauskas
https://arxiv.org/abs/2404.07715
This https://arxiv.org/abs/2404.05426 has been replaced.
initial toot: https://mastoxiv.page/@arXiv_csCV_…
'One size doesn't fit all': Learning how many Examples to use for In-Context Learning for Improved Text Classification
Manish Chandra, Debasis Ganguly, Yiwen Li, Iadh Ounis
https://arxiv.org/abs/2403.06402
Jack Clark’s Import AI newsletter is always fun; this one especially because it explores the shared interest Meta and the CCP have in preventing fine-tuning of their models.
In a “Palantír”-level irony, one of the projects to prevent all further learning is called SOPHON. https://
Data-driven sparse modeling of oscillations in plasma space propulsion
B. Bay\'on-Buj\'an, M. Merino
https://arxiv.org/abs/2403.06809 https://
SAM-I-Am: Semantic Boosting for Zero-shot Atomic-Scale Electron Micrograph Segmentation
Waqwoya Abebe, Jan Strube, Luanzheng Guo, Nathan R. Tallent, Oceane Bel, Steven Spurgeon, Christina Doty, Ali Jannesari
https://arxiv.org/abs/2404.06638
Geometry from geodesics: fine-tuning Ehlers, Pirani, and Schild
James T. Wheeler
https://arxiv.org/abs/2404.03815 https://arxiv.org/p…
OpenAI expands its Custom Model training program with "assisted fine-tuning", letting organizations set up data training pipelines, evaluation systems, and more (Kyle Wiggers/TechCrunch)
https://techcrunch.com/2024/04/04/openai-expands…
Aya Dataset: An Open-Access Collection for Multilingual Instruction Tuning
Shivalika Singh, Freddie Vargus, Daniel Dsouza, B\"orje F. Karlsson, Abinaya Mahendiran, Wei-Yin Ko, Herumb Shandilya, Jay Patel, Deividas Mataciunas, Laura OMahony, Mike Zhang, Ramith Hettiarachchi, Joseph Wilson, Marina Machado, Luisa Souza Moura, Dominik Krzemi\'nski, Hakimeh Fadaei, Irem Erg\"un, Ifeoma Okoh, Aisha Alaagib, Oshan Mudannayake, Zaid Alyafeai, Vu Minh Chien, Sebastian Ruder, Surya…
Speaking of vaccines, this could be very important. https://mag.uchicago.edu/science-medicine/fine-tuning-immunity
This https://arxiv.org/abs/2402.04004 has been replaced.
link: https://scholar.google.com/scholar?q=a
This https://arxiv.org/abs/2310.06611 has been replaced.
initial toot: https://mastoxiv.page/@arXiv_hepp…
This https://arxiv.org/abs/2312.03045 has been replaced.
initial toot: https://mastoxiv.page/@arXiv_csCV_…
Low-Dose CT Image Reconstruction by Fine-Tuning a UNet Pretrained for Gaussian Denoising for the Downstream Task of Image Enhancement
Tim Selig, Thomas M\"arz, Martin Storath, Andreas Weinmann
https://arxiv.org/abs/2403.03551
AD-NEv : The multi-architecture neuroevolution-based multivariate anomaly detection framework
Marcin Pietro\'n, Dominik \.Zurek, Kamil Faber, Roberto Corizzo
https://arxiv.org/abs/2404.07968
This https://arxiv.org/abs/2307.07218 has been replaced.
initial toot: https://mastoxiv.page/@arXiv_ees…
This https://arxiv.org/abs/2402.10100 has been replaced.
initial toot: https://mastoxiv.page/@arXiv_csSD_…
Iris-SAM: Iris Segmentation Using a Foundational Model
Parisa Farmanifard, Arun Ross
https://arxiv.org/abs/2402.06497 https://arxiv.o…
This https://arxiv.org/abs/2401.04190 has been replaced.
initial toot: https://mastoxiv.page/@arXiv_…
This https://arxiv.org/abs/2305.00418 has been replaced.
initial toot: https://mastoxiv.page/@arXiv_csSE_…
This https://arxiv.org/abs/2403.16915 has been replaced.
initial toot: https://mastoxiv.page/@arXiv_csIR_…
PLLM-CS: Pre-trained Large Language Model (LLM) for Cyber Threat Detection in Satellite Networks
Mohammed Hassanin, Marwa Keshk, Sara Salim, Majid Alsubaie, Dharmendra Sharma
https://arxiv.org/abs/2405.05469
Wiley licenses content for training an #LLM. The company was not named, but I would suspect it's the one which has been signing a lot of licensing deals lately. Access to STM content could be a big differentiator, though I wouldn't expect it to be exclusive. Also, $23M sounds small.
This https://arxiv.org/abs/2310.05910 has been replaced.
link: https://scholar.google.com/scholar?q=a
This https://arxiv.org/abs/2405.02422 has been replaced.
initial toot: https://mastoxiv.page/@arXiv_ees…
Higgs Alignment from Multicritical-Point Principle in Two Higgs Doublet Models
Hikaru Kawai, Kiyoharu Kawana, Kin-ya Oda, Kei Yagyu
https://arxiv.org/abs/2404.06096
This https://arxiv.org/abs/2312.09979 has been replaced.
initial toot: https://mastoxiv.page/@arXiv_csCL_…
This https://arxiv.org/abs/2306.07285 has been replaced.
initial toot: https://mastoxiv.page/@arXiv_csSE_…
Deep Prompt Multi-task Network for Abuse Language Detection
Jian Zhu, Yuping Ruan, Jingfei Chang, Cheng Luo
https://arxiv.org/abs/2403.05268 https://
This https://arxiv.org/abs/2401.04190 has been replaced.
initial toot: https://mastoxiv.page/@arXiv_…
This https://arxiv.org/abs/2404.15786 has been replaced.
link: https://scholar.google.com/scholar?q=a
Magic-Boost: Boost 3D Generation with Mutli-View Conditioned Diffusion
Fan Yang, Jianfeng Zhang, Yichun Shi, Bowen Chen, Chenxu Zhang, Huichao Zhang, Xiaofeng Yang, Jiashi Feng, Guosheng Lin
https://arxiv.org/abs/2404.06429
Deep Prompt Multi-task Network for Abuse Language Detection
Jian Zhu, Yuping Ruan, Jingfei Chang, Cheng Luo
https://arxiv.org/abs/2403.05268 https://
Coarse-Tuning for Ad-hoc Document Retrieval Using Pre-trained Language Models
Atsushi Keyaki, Ribeka Keyaki
https://arxiv.org/abs/2403.16915 https://
Cross-lingual Transfer or Machine Translation? On Data Augmentation for Monolingual Semantic Textual Similarity
Sho Hoshino, Akihiko Kato, Soichiro Murakami, Peinan Zhang
https://arxiv.org/abs/2403.05257
Mitigating Fine-tuning Jailbreak Attack with Backdoor Enhanced Alignment
Jiongxiao Wang, Jiazhao Li, Yiquan Li, Xiangyu Qi, Muhao Chen, Junjie Hu, Yixuan Li, Bo Li, Chaowei Xiao
https://arxiv.org/abs/2402.14968
Cross-lingual Transfer or Machine Translation? On Data Augmentation for Monolingual Semantic Textual Similarity
Sho Hoshino, Akihiko Kato, Soichiro Murakami, Peinan Zhang
https://arxiv.org/abs/2403.05257
Can Feedback Enhance Semantic Grounding in Large Vision-Language Models?
Yuan-Hong Liao, Rafid Mahmood, Sanja Fidler, David Acuna
https://arxiv.org/abs/2404.06510
This https://arxiv.org/abs/2304.04067 has been replaced.
initial toot: https://mastoxiv.page/@arXiv_csNE_…
This https://arxiv.org/abs/2305.18582 has been replaced.
initial toot: https://mastoxiv.page/@arXiv_csCL_…
This https://arxiv.org/abs/2305.18582 has been replaced.
initial toot: https://mastoxiv.page/@arXiv_csCL_…
FT2Ra: A Fine-Tuning-Inspired Approach to Retrieval-Augmented Code Completion
Qi Guo, Xiaohong Li, Xiaofei Xie, Shangqing Liu, Ze Tang, Ruitao Feng, Junjie Wang, Jidong Ge, Lei Bu
https://arxiv.org/abs/2404.01554
This https://arxiv.org/abs/2404.11536 has been replaced.
link: https://scholar.google.com/scholar?q=a
This https://arxiv.org/abs/2402.14968 has been replaced.
initial toot: https://mastoxiv.page/@arXiv_csCR_…
This https://arxiv.org/abs/2403.16915 has been replaced.
initial toot: https://mastoxiv.page/@arXiv_csIR_…
Bias-Augmented Consistency Training Reduces Biased Reasoning in Chain-of-Thought
James Chua, Edward Rees, Hunar Batra, Samuel R. Bowman, Julian Michael, Ethan Perez, Miles Turpin
https://arxiv.org/abs/2403.05518
Bias-Augmented Consistency Training Reduces Biased Reasoning in Chain-of-Thought
James Chua, Edward Rees, Hunar Batra, Samuel R. Bowman, Julian Michael, Ethan Perez, Miles Turpin
https://arxiv.org/abs/2403.05518
This https://arxiv.org/abs/2305.13179 has been replaced.
link: https://scholar.google.com/scholar?q=a
Learning-to-learn enables rapid learning with phase-change memory-based in-memory computing
Thomas Ortner, Horst Petschenig, Athanasios Vasilopoulos, Roland Renner, \v{S}pela Brglez, Thomas Limbacher, Enrique Pi\~nero, Alejandro Linares Barranco, Angeliki Pantazi, Robert Legenstein
https://arxiv.org/abs/2405.05141
Empirical Studies of Parameter Efficient Methods for Large Language Models of Code and Knowledge Transfer to R
Amirreza Esmaeili, Iman Saberi, Fatemeh H. Fard
https://arxiv.org/abs/2405.01553
This https://arxiv.org/abs/2305.13179 has been replaced.
link: https://scholar.google.com/scholar?q=a
Unlocking Parameter-Efficient Fine-Tuning for Low-Resource Language Translation
Tong Su, Xin Peng, Sarubi Thillainathan, David Guzm\'an, Surangika Ranathunga, En-Shiun Annie Lee
https://arxiv.org/abs/2404.04212
A Fine-tuning Enhanced RAG System with Quantized Influence Measure as AI Judge
Keshav Rangan, Yiqiao Yin
https://arxiv.org/abs/2402.17081 https://
This https://arxiv.org/abs/2402.12168 has been replaced.
initial toot: https://mastoxiv.page/@arXiv_csCR_…
This https://arxiv.org/abs/2404.14367 has been replaced.
initial toot: https://mastoxiv.page/@arXiv_csLG_…
Fine-tuning Pre-trained Named Entity Recognition Models For Indian Languages
Sankalp Bahad, Pruthwik Mishra, Karunesh Arora, Rakesh Chandra Balabantaray, Dipti Misra Sharma, Parameswari Krishnamurthy
https://arxiv.org/abs/2405.04829
Fine-tuning Pre-trained Named Entity Recognition Models For Indian Languages
Sankalp Bahad, Pruthwik Mishra, Karunesh Arora, Rakesh Chandra Balabantaray, Dipti Misra Sharma, Parameswari Krishnamurthy
https://arxiv.org/abs/2405.04829
Constraining Large Language Model for Generating Computer-Parsable Content
Jiaye Wang
https://arxiv.org/abs/2404.05499 https://arxiv.…
This https://arxiv.org/abs/2310.09266 has been replaced.
initial toot: https://mastoxiv.page/@arXiv_csCR_…
This https://arxiv.org/abs/2402.17412 has been replaced.
initial toot: https://mastoxiv.page/@arXiv_csCV_…
This https://arxiv.org/abs/2403.01432 has been replaced.
link: https://scholar.google.com/scholar?q=a
This https://arxiv.org/abs/2402.00905 has been replaced.
link: https://scholar.google.com/scholar?q=a
LoRA Land: 310 Fine-tuned LLMs that Rival GPT-4, A Technical Report
Justin Zhao, Timothy Wang, Wael Abid, Geoffrey Angus, Arnav Garg, Jeffery Kinnison, Alex Sherstinsky, Piero Molino, Travis Addair, Devvret Rishi
https://arxiv.org/abs/2405.00732
Multi-Objective Fine-Tuning for Enhanced Program Repair with LLMs
Boyang Yang, Haoye Tian, Jiadong Ren, Hongyu Zhang, Jacques Klein, Tegawend\'e F. Bissyand\'e, Claire Le Goues, Shunfu Jin
https://arxiv.org/abs/2404.12636
This https://arxiv.org/abs/2402.00905 has been replaced.
link: https://scholar.google.com/scholar?q=a
This https://arxiv.org/abs/2403.20145 has been replaced.
initial toot: https://mastoxiv.page/@arXiv_csCL_…
Fine-Tuning Enhances Existing Mechanisms: A Case Study on Entity Tracking
Nikhil Prakash, Tamar Rott Shaham, Tal Haklay, Yonatan Belinkov, David Bau
https://arxiv.org/abs/2402.14811
This https://arxiv.org/abs/2403.09891 has been replaced.
initial toot: https://mastoxiv.page/@arXiv_csCL_…
OpenMedLM: Prompt engineering can out-perform fine-tuning in medical question-answering with open-source large language models
Jenish Maharjan, Anurag Garikipati, Navan Preet Singh, Leo Cyrus, Mayank Sharma, Madalina Ciobanu, Gina Barnes, Rahul Thapa, Qingqing Mao, Ritankar Das
https://arxiv.org/abs/2402.19371
Cross-Lingual Learning vs. Low-Resource Fine-Tuning: A Case Study with Fact-Checking in Turkish
Recep Firat Cekinel, Pinar Karagoz, Cagri Coltekin
https://arxiv.org/abs/2403.00411
Learning or Self-aligning? Rethinking Instruction Fine-tuning
Mengjie Ren, Boxi Cao, Hongyu Lin, Liu Cao, Xianpei Han, Ke Zeng, Guanglu Wan, Xunliang Cai, Le Sun
https://arxiv.org/abs/2402.18243
This https://arxiv.org/abs/2404.00213 has been replaced.
link: https://scholar.google.com/scholar?q=a
Simple Techniques for Enhancing Sentence Embeddings in Generative Language Models
Bowen Zhang, Kehua Chang, Chunping Li
https://arxiv.org/abs/2404.03921 ht…
This https://arxiv.org/abs/2309.13734 has been replaced.
link: https://scholar.google.com/scholar?q=a
Is Crowdsourcing Breaking Your Bank? Cost-Effective Fine-Tuning of Pre-trained Language Models with Proximal Policy Optimization
Shuo Yang, Gjergji Kasneci
https://arxiv.org/abs/2402.18284
This https://arxiv.org/abs/2403.18025 has been replaced.
initial toot: https://mastoxiv.page/@arXiv_csCL_…
INSTRAUG: Automatic Instruction Augmentation for Multimodal Instruction Fine-tuning
Wei Han, Hui Chen, Soujanya Poria
https://arxiv.org/abs/2402.14492 http…